Conference Proceedings

KDGAN: Knowledge distillation with generative adversarial networks

X Wang, Y Sun, R Zhang, J Qi

Advances in Neural Information Processing Systems | NEURAL INFORMATION PROCESSING SYSTEMS (NIPS) | Published : 2018

Abstract

Knowledge distillation (KD) aims to train a lightweight classifier suitable to provide accurate inference with constrained resources in multi-label learning. Instead of directly consuming feature-label pairs, the classifier is trained by a teacher, i.e., a high-capacity model whose training may be resource-hungry. The accuracy of the classifier trained this way is usually suboptimal because it is difficult to learn the true data distribution from the teacher. An alternative method is to adversarially train the classifier against a discriminator in a two-player game akin to generative adversarial networks (GAN), which can ensure the classifier to learn the true data distribution at the equili..

View full abstract

University of Melbourne Researchers

Grants

Awarded by Australian Research Council


Funding Acknowledgements

This work is supported by Australian Research Council Future Fellowship Project FT120100832 and Discovery Project DP180102050. We thank the anonymous reviewers for their feedback on the paper. We have incorporated responses to reviewers' comments in the paper.